# Lightweight Language Model
Minillm 0.2B WithWudao
Apache-2.0
MiniLLM is a lightweight Chinese language model developed based on the bert4torch framework, covering the entire process from pre-training to instruction fine-tuning, with basic dialogue capabilities
Large Language Model
Transformers

M
Tongjilibo
127
2
Latent Recurrent Depth Lm
MIT
An experimental text generation architecture that captures deeper contextual information through iterative latent processing
Large Language Model
Transformers English

L
codewithdark
38
1
Miniplm Qwen 200M
Apache-2.0
A 200M-parameter model based on the Qwen architecture, pretrained from scratch using the MiniPLM knowledge distillation framework
Large Language Model
Transformers English

M
MiniLLM
203
5
Meta Llama 3.1 8B Instruct Abliterated GGUF
MIT
A text generation model employing mixed quantization techniques, with output and embedding tensors in f16 format and other tensors quantized using q5_k or q6_k. It has a smaller size than the standard q8_0 quantization format while maintaining performance comparable to the pure f16 version.
Large Language Model English
M
ZeroWw
98
17
Mamba 3B Slimpj
Apache-2.0
A 3B-parameter language model based on the Mamba architecture, supporting English text generation tasks.
Large Language Model
Transformers English

M
Q-bert
56
3
Llama2 Xs 460M Experimental
This series of repositories open-sources reproductions of Meta AI's LLaMA and LLaMA 2 large language models, but with significantly reduced model sizes. The llama1_s experimental version contains 1.8 billion parameters, while the llama2_xs experimental version has only 460 million parameters.
Large Language Model
Transformers English

L
ahxt
145
13
Gpt2023
MIT
A 124M-parameter language model based on the GPT-2 architecture, fine-tuned on 2.23B tokens of diverse data with improved text generation capabilities
Large Language Model
Transformers English

G
crumb
136
18
Japanese Gpt Neox Small
MIT
A small Japanese language model based on GPT-NeoX architecture, supporting text generation tasks
Large Language Model
Transformers Supports Multiple Languages

J
rinna
838
15
Albert Base Japanese V1 With Japanese Tokenizer
MIT
This is a Japanese-pretrained ALBERT model that uses BertJapaneseTokenizer as its tokenizer, making Japanese text processing more convenient.
Large Language Model
Transformers Japanese

A
ken11
44
3
Mminilmv2 L6 H384 Distilled From XLMR Large
MiniLMv2 is a lightweight language representation model developed by Microsoft, achieving efficient performance through knowledge distillation technology.
Large Language Model
Transformers

M
nreimers
197
17
Bert L12 H384 A6
A lightweight BERT model pre-trained on the BookCorpus dataset using knowledge distillation technology, with the hidden layer dimension reduced to 384 and 6 attention heads.
Large Language Model
Transformers

B
eli4s
16
2
Distilbert Base Uncased Sparse 90 Unstructured Pruneofa
Apache-2.0
This is a sparse pre-trained model achieving 90% weight sparsity through one-shot pruning, suitable for fine-tuning on various language tasks.
Large Language Model
Transformers English

D
Intel
78
2
Featured Recommended AI Models